25 research outputs found

    AI-based Sound-Squatting Attack Made Possible

    Get PDF
    Domain squatting is an efficient attacking technique that relies on the similarity between domain names to trick users. Sound-squatting is a type of domain squatting that explores the similarity in the pronunciation of domains. Sound-squatting requires better approaches to protect users, and indeed it demands more research attention due to popularization of intelligent speakers and the increase of voice-based navigation. In this work we propose an AI-based methodology to automatically build sound-squatting candidates. We leverage recent results of AI, namely the ability to translate text, to automatically generate possible sound-squatting candidates. We evaluate our methodology by verifying the generated candidates and classifying them according to their threat class. We generate over twenty thousand candidates from popular domains, out of which, 7% are found active at the time of the analysis. Active domains include “Parked/Ads/For-Sale” domains. We thus show that automatic sound-squatting generation is useful to proactively check and limit the abuse of such offence

    Tracking Knowledge Propagation Across Wikipedia Languages

    Get PDF
    In this paper, we present a dataset of inter-language knowledge propagation in Wikipedia. Covering the entire 309 language editions and 33M articles, the dataset aims to track the full propagation history of Wikipedia concepts, and allow follow-up research on building predictive models of them. For this purpose, we align all the Wikipedia articles in a language-agnostic manner according to the concept they cover, which results in 13M propagation instances. To the best of our knowledge, this dataset is the first to explore the full inter-language propagation at a large scale. Together with the dataset, a holistic overview of the propagation and key insights about the underlying structural factors are provided to aid future research. For example, we find that although long cascades are unusual, the propagation tends to continue further once it reaches more than four language editions. We also find that the size of language editions is associated with the speed of propagation. We believe the dataset not only contributes to the prior literature on Wikipedia growth but also enables new use cases such as edit recommendation for addressing knowledge gaps, detection of disinformation, and cultural relationship analysis

    Basin-wide variation in tree hydraulic safety margins predicts the carbon balance of Amazon forests

    Get PDF
    Funding: Data collection was largely funded by the UK Natural Environment Research Council (NERC) project TREMOR (NE/N004655/1) to D.G., E.G. and O.P., with further funds from Coordenação de Aperfeiçoamento de Pessoal de NĂ­vel Superior—Brasil (CAPES, finance code 001) to J.V.T. and a University of Leeds Climate Research Bursary Fund to J.V.T. D.G., E.G. and O.P. acknowledge further support from a NERC-funded consortium award (ARBOLES, NE/S011811/1). This paper is an outcome of J.V.T.’s doctoral thesis, which was sponsored by CAPES (GDE 99999.001293/2015-00). J.V.T. was previously supported by the NERC-funded ARBOLES project (NE/S011811/1) and is supported at present by the Swedish Research Council VetenskapsrĂ„det (grant no. 2019-03758 to R.M.). E.G., O.P. and D.G. acknowledge support from NERC-funded BIORED grant (NE/N012542/1). O.P. acknowledges support from an ERC Advanced Grant and a Royal Society Wolfson Research Merit Award. R.S.O. was supported by a CNPq productivity scholarship, the SĂŁo Paulo Research Foundation (FAPESP-Microsoft 11/52072-0) and the US Department of Energy, project GoAmazon (FAPESP 2013/50531-2). M.M. acknowledges support from MINECO FUN2FUN (CGL2013-46808-R) and DRESS (CGL2017-89149-C2-1-R). C.S.-M., F.B.V. and P.R.L.B. were financed by Coordenação de Aperfeiçoamento de Pessoal de NĂ­vel Superior—Brasil (CAPES, finance code 001). C.S.-M. received a scholarship from the Brazilian National Council for Scientific and Technological Development (CNPq 140353/2017-8) and CAPES (science without borders 88881.135316/2016-01). Y.M. acknowledges the Gordon and Betty Moore Foundation and ERC Advanced Investigator Grant (GEM-TRAITS, 321131) for supporting the Global Ecosystems Monitoring (GEM) network (gem.tropicalforests.ox.ac.uk), within which some of the field sites (KEN, TAM and ALP) are nested. The authors thank Brazil–USA Collaborative Research GoAmazon DOE-FAPESP-FAPEAM (FAPESP 2013/50533-5 to L.A.) and National Science Foundation (award DEB-1753973 to L. Alves). They thank Serrapilheira Serra-1709-18983 (to M.H.) and CNPq-PELD/POPA-441443/2016-8 (to L.G.) (P.I. Albertina Lima). They thank all the colleagues and grants mentioned elsewhere [8,36] that established, identified and measured the Amazon forest plots in the RAINFOR network analysed here. The authors particularly thank J. Lyod, S. Almeida, F. Brown, B. Vicenti, N. Silva and L. Alves. This work is an outcome approved Research Project no. 19 from ForestPlots.net, a collaborative initiative developed at the University of Leeds that unites researchers and the monitoring of their permanent plots from the world’s tropical forests [61]. The authros thank A. Levesley, K. Melgaço Ladvocat and G. Pickavance for ForestPlots.net management. They thank Y. Wang and J. Baker, respectively, for their help with the map and with the climatic data. The authors acknowledge the invaluable help of M. Brum for kindly providing the comparison of vulnerability curves based on PAD and on PLC shown in this manuscript. They thank J. Martinez-Vilalta for his comments on an early version of this manuscript. The authors also thank V. Hilares and the AsociaciĂłn para la InvestigaciĂłn y Desarrollo Integral (AIDER, Puerto Maldonado, Peru); V. Saldaña and Instituto de Investigaciones de la AmazonĂ­a Peruana (IIAP) for local field campaign support in Peru; E. Chavez and Noel Kempff Natural History Museum for local field campaign support in Bolivia; ICMBio, INPA/NAPPA/LBA COOMFLONA (Cooperativa mista da Flona TapajĂłs) and T. I. Bragança-Marituba for the research support.Tropical forests face increasing climate risk1,2, yet our ability to predict their response to climate change is limited by poor understanding of their resistance to water stress. Although xylem embolism resistance thresholds (for example, Κ50) and hydraulic safety margins (for example, HSM50) are important predictors of drought-induced mortality risk3-5, little is known about how these vary across Earth's largest tropical forest. Here, we present a pan-Amazon, fully standardized hydraulic traits dataset and use it to assess regional variation in drought sensitivity and hydraulic trait ability to predict species distributions and long-term forest biomass accumulation. Parameters Κ50 and HSM50 vary markedly across the Amazon and are related to average long-term rainfall characteristics. Both Κ50 and HSM50 influence the biogeographical distribution of Amazon tree species. However, HSM50 was the only significant predictor of observed decadal-scale changes in forest biomass. Old-growth forests with wide HSM50 are gaining more biomass than are low HSM50 forests. We propose that this may be associated with a growth-mortality trade-off whereby trees in forests consisting of fast-growing species take greater hydraulic risks and face greater mortality risk. Moreover, in regions of more pronounced climatic change, we find evidence that forests are losing biomass, suggesting that species in these regions may be operating beyond their hydraulic limits. Continued climate change is likely to further reduce HSM50 in the Amazon6,7, with strong implications for the Amazon carbon sink.Publisher PDFPeer reviewe

    Highly-parallelized simulation of a pixelated LArTPC on a GPU

    No full text
    The rapid development of general-purpose computing on graphics processing units (GPGPU) is allowing the implementation of highly-parallelized Monte Carlo simulation chains for particle physics experiments. This technique is particularly suitable for the simulation of a pixelated charge readout for time projection chambers, given the large number of channels that this technology employs. Here we present the first implementation of a full microphysical simulator of a liquid argon time projection chamber (LArTPC) equipped with light readout and pixelated charge readout, developed for the DUNE Near Detector. The software is implemented with an end-to-end set of GPU-optimized algorithms. The algorithms have been written in Python and translated into CUDA kernels using Numba, a just-in-time compiler for a subset of Python and NumPy instructions. The GPU implementation achieves a speed up of four orders of magnitude compared with the equivalent CPU version. The simulation of the current induced on 10310^3 pixels takes around 1 ms on the GPU, compared with approximately 10 s on the CPU. The results of the simulation are compared against data from a pixel-readout LArTPC prototype

    DUNE Offline Computing Conceptual Design Report

    No full text
    This document describes Offline Software and Computing for the Deep Underground Neutrino Experiment (DUNE) experiment, in particular, the conceptual design of the offline computing needed to accomplish its physics goals. Our emphasis in this document is the development of the computing infrastructure needed to acquire, catalog, reconstruct, simulate and analyze the data from the DUNE experiment and its prototypes. In this effort, we concentrate on developing the tools and systems that facilitate the development and deployment of advanced algorithms. Rather than prescribing particular algorithms, our goal is to provide resources that are flexible and accessible enough to support creative software solutions as HEP computing evolves and to provide computing that achieves the physics goals of the DUNE experiment.This document describes the conceptual design for the Offline Software and Computing for the Deep Underground Neutrino Experiment (DUNE). The goals of the experiment include 1) studying neutrino oscillations using a beam of neutrinos sent from Fermilab in Illinois to the Sanford Underground Research Facility (SURF) in Lead, South Dakota, 2) studying astrophysical neutrino sources and rare processes and 3) understanding the physics of neutrino interactions in matter. We describe the development of the computing infrastructure needed to achieve the physics goals of the experiment by storing, cataloging, reconstructing, simulating, and analyzing ∌\sim 30 PB of data/year from DUNE and its prototypes. Rather than prescribing particular algorithms, our goal is to provide resources that are flexible and accessible enough to support creative software solutions and advanced algorithms as HEP computing evolves. We describe the physics objectives, organization, use cases, and proposed technical solutions

    DUNE Offline Computing Conceptual Design Report

    No full text
    This document describes Offline Software and Computing for the Deep Underground Neutrino Experiment (DUNE) experiment, in particular, the conceptual design of the offline computing needed to accomplish its physics goals. Our emphasis in this document is the development of the computing infrastructure needed to acquire, catalog, reconstruct, simulate and analyze the data from the DUNE experiment and its prototypes. In this effort, we concentrate on developing the tools and systems thatfacilitate the development and deployment of advanced algorithms. Rather than prescribing particular algorithms, our goal is to provide resources that are flexible and accessible enough to support creative software solutions as HEP computing evolves and to provide computing that achieves the physics goals of the DUNE experiment

    Reconstruction of interactions in the ProtoDUNE-SP detector with Pandora

    No full text
    International audienceThe Pandora Software Development Kit and algorithm libraries provide pattern-recognition logic essential to the reconstruction of particle interactions in liquid argon time projection chamber detectors. Pandora is the primary event reconstruction software used at ProtoDUNE-SP, a prototype for the Deep Underground Neutrino Experiment far detector. ProtoDUNE-SP, located at CERN, is exposed to a charged-particle test beam. This paper gives an overview of the Pandora reconstruction algorithms and how they have been tailored for use at ProtoDUNE-SP. In complex events with numerous cosmic-ray and beam background particles, the simulated reconstruction and identification efficiency for triggered test-beam particles is above 80% for the majority of particle type and beam momentum combinations. Specifically, simulated 1 GeV/cc charged pions and protons are correctly reconstructed and identified with efficiencies of 86.1±0.6\pm0.6% and 84.1±0.6\pm0.6%, respectively. The efficiencies measured for test-beam data are shown to be within 5% of those predicted by the simulation

    Reconstruction of interactions in the ProtoDUNE-SP detector with Pandora

    No full text
    International audienceThe Pandora Software Development Kit and algorithm libraries provide pattern-recognition logic essential to the reconstruction of particle interactions in liquid argon time projection chamber detectors. Pandora is the primary event reconstruction software used at ProtoDUNE-SP, a prototype for the Deep Underground Neutrino Experiment far detector. ProtoDUNE-SP, located at CERN, is exposed to a charged-particle test beam. This paper gives an overview of the Pandora reconstruction algorithms and how they have been tailored for use at ProtoDUNE-SP. In complex events with numerous cosmic-ray and beam background particles, the simulated reconstruction and identification efficiency for triggered test-beam particles is above 80% for the majority of particle type and beam momentum combinations. Specifically, simulated 1 GeV/cc charged pions and protons are correctly reconstructed and identified with efficiencies of 86.1±0.6\pm0.6% and 84.1±0.6\pm0.6%, respectively. The efficiencies measured for test-beam data are shown to be within 5% of those predicted by the simulation

    Separation of track- and shower-like energy deposits in ProtoDUNE-SP using a convolutional neural network

    No full text
    International audienceLiquid argon time projection chamber detector technology provides high spatial and calorimetric resolutions on the charged particles traversing liquid argon. As a result, the technology has been used in a number of recent neutrino experiments, and is the technology of choice for the Deep Underground Neutrino Experiment (DUNE). In order to perform high precision measurements of neutrinos in the detector, final state particles need to be effectively identified, and their energy accurately reconstructed. This article proposes an algorithm based on a convolutional neural network to perform the classification of energy deposits and reconstructed particles as track-like or arising from electromagnetic cascades. Results from testing the algorithm on experimental data from ProtoDUNE-SP, a prototype of the DUNE far detector, are presented. The network identifies track- and shower-like particles, as well as Michel electrons, with high efficiency. The performance of the algorithm is consistent between experimental data and simulation

    Highly-parallelized simulation of a pixelated LArTPC on a GPU

    No full text
    The rapid development of general-purpose computing on graphics processing units (GPGPU) is allowing the implementation of highly-parallelized Monte Carlo simulation chains for particle physics experiments. This technique is particularly suitable for the simulation of a pixelated charge readout for time projection chambers, given the large number of channels that this technology employs. Here we present the first implementation of a full microphysical simulator of a liquid argon time projection chamber (LArTPC) equipped with light readout and pixelated charge readout, developed for the DUNE Near Detector. The software is implemented with an end-to-end set of GPU-optimized algorithms. The algorithms have been written in Python and translated into CUDA kernels using Numba, a just-in-time compiler for a subset of Python and NumPy instructions. The GPU implementation achieves a speed up of four orders of magnitude compared with the equivalent CPU version. The simulation of the current induced on 10310^3 pixels takes around 1 ms on the GPU, compared with approximately 10 s on the CPU. The results of the simulation are compared against data from a pixel-readout LArTPC prototype
    corecore